Enhancing Parameter-Free Frank Wolfe with an Extra Subproblem

نویسندگان

چکیده

Aiming at convex optimization under structural constraints, this work introduces and analyzes a variant of the Frank Wolfe (FW) algorithm termed ExtraFW. The distinct feature ExtraFW is pair gradients leveraged per iteration, thanks to which decision variable updated in prediction-correction (PC) format. Relying on no problem dependent parameters step sizes, convergence rate for general problems shown be ${\cal O}(\frac{1}{k})$, optimal sense matching lower bound number solved FW subproblems. However, merit its faster O}\big(\frac{1}{k^2} \big)$ class machine learning problems. Compared with other parameter-free variants that have rates same problems, has improved fine-grained analysis PC update. Numerical tests binary classification different sparsity-promoting constraints demonstrate empirical performance significantly better than FW, even Nesterov's accelerated gradient certain datasets. For matrix completion, enjoys smaller optimality gap, rank FW.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Frank-Wolfe with Subsampling Oracle

We analyze two novel randomized variants of the Frank-Wolfe (FW) or conditional gradient algorithm. While classical FW algorithms require solving a linear minimization problem over the domain at each iteration, the proposedmethod only requires to solve a linear minimization problem over a small subset of the original domain. The first algorithm that we propose is a randomized variant of the ori...

متن کامل

Revisiting Frank-Wolfe: Projection-Free Sparse Convex Optimization

We provide stronger and more general primal-dual convergence results for FrankWolfe-type algorithms (a.k.a. conditional gradient) for constrained convex optimization, enabled by a simple framework of duality gap certificates. Our analysis also holds if the linear subproblems are only solved approximately (as well as if the gradients are inexact), and is proven to be worst-case optimal in the sp...

متن کامل

Combining Progressive Hedging with a Frank-wolfe

We present a new primal-dual algorithm for computing the value of the Lagrangian 6 dual of a stochastic mixed-integer program (SMIP) formed by relaxing its nonanticipativity con7 straints. The algorithm relies on the well-known progressive hedging method, but unlike previous 8 progressive hedging approaches for SMIP, our algorithm can be shown to converge to the optimal 9 Lagrangian dual value....

متن کامل

Learning Infinite RBMs with Frank-Wolfe

In this work, we propose an infinite restricted Boltzmann machine (RBM), whose maximum likelihood estimation (MLE) corresponds to a constrained convex optimization. We consider the Frank-Wolfe algorithm to solve the program, which provides a sparse solution that can be interpreted as inserting a hidden unit at each iteration, so that the optimization process takes the form of a sequence of fini...

متن کامل

Bandit Optimization with Upper-Confidence Frank-Wolfe

We consider the problem of bandit optimization, inspired by stochastic optimization and online learning problems with bandit feedback. In this problem, the objective is to minimize a global loss function of all the actions, not necessarily a cumulative loss. This framework allows us to study a very general class of problems, with applications in statistics, machine learning, and other fields. T...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Proceedings of the ... AAAI Conference on Artificial Intelligence

سال: 2021

ISSN: ['2159-5399', '2374-3468']

DOI: https://doi.org/10.1609/aaai.v35i9.17012